40 research outputs found
The Science of Detecting LLM-Generated Texts
The emergence of large language models (LLMs) has resulted in the production
of LLM-generated texts that is highly sophisticated and almost
indistinguishable from texts written by humans. However, this has also sparked
concerns about the potential misuse of such texts, such as spreading
misinformation and causing disruptions in the education system. Although many
detection approaches have been proposed, a comprehensive understanding of the
achievements and challenges is still lacking. This survey aims to provide an
overview of existing LLM-generated text detection techniques and enhance the
control and regulation of language generation models. Furthermore, we emphasize
crucial considerations for future research, including the development of
comprehensive evaluation metrics and the threat posed by open-source LLMs, to
drive progress in the area of LLM-generated text detection
SPeC: A Soft Prompt-Based Calibration on Mitigating Performance Variability in Clinical Notes Summarization
Electronic health records (EHRs) store an extensive array of patient
information, encompassing medical histories, diagnoses, treatments, and test
outcomes. These records are crucial for enabling healthcare providers to make
well-informed decisions regarding patient care. Summarizing clinical notes
further assists healthcare professionals in pinpointing potential health risks
and making better-informed decisions. This process contributes to reducing
errors and enhancing patient outcomes by ensuring providers have access to the
most pertinent and current patient data. Recent research has shown that
incorporating prompts with large language models (LLMs) substantially boosts
the efficacy of summarization tasks. However, we show that this approach also
leads to increased output variance, resulting in notably divergent outputs even
when prompts share similar meanings. To tackle this challenge, we introduce a
model-agnostic Soft Prompt-Based Calibration (SPeC) pipeline that employs soft
prompts to diminish variance while preserving the advantages of prompt-based
summarization. Experimental findings on multiple clinical note tasks and LLMs
indicate that our method not only bolsters performance but also effectively
curbs variance for various LLMs, providing a more uniform and dependable
solution for summarizing vital medical information
DISPEL: Domain Generalization via Domain-Specific Liberating
Domain generalization aims to learn a generalization model that can perform
well on unseen test domains by only training on limited source domains.
However, existing domain generalization approaches often bring in
prediction-irrelevant noise or require the collection of domain labels. To
address these challenges, we consider the domain generalization problem from a
different perspective by categorizing underlying feature groups into
domain-shared and domain-specific features. Nevertheless, the domain-specific
features are difficult to be identified and distinguished from the input data.
In this work, we propose DomaIn-SPEcific Liberating (DISPEL), a post-processing
fine-grained masking approach that can filter out undefined and
indistinguishable domain-specific features in the embedding space.
Specifically, DISPEL utilizes a mask generator that produces a unique mask for
each input data to filter domain-specific features. The DISPEL framework is
highly flexible to be applied to any fine-tuned models. We derive a
generalization error bound to guarantee the generalization performance by
optimizing a designed objective loss. The experimental results on five
benchmarks demonstrate DISPEL outperforms existing methods and can further
generalize various algorithms
Efficient XAI Techniques: A Taxonomic Survey
Recently, there has been a growing demand for the deployment of Explainable
Artificial Intelligence (XAI) algorithms in real-world applications. However,
traditional XAI methods typically suffer from a high computational complexity
problem, which discourages the deployment of real-time systems to meet the
time-demanding requirements of real-world scenarios. Although many approaches
have been proposed to improve the efficiency of XAI methods, a comprehensive
understanding of the achievements and challenges is still needed. To this end,
in this paper we provide a review of efficient XAI. Specifically, we categorize
existing techniques of XAI acceleration into efficient non-amortized and
efficient amortized methods. The efficient non-amortized methods focus on
data-centric or model-centric acceleration upon each individual instance. In
contrast, amortized methods focus on learning a unified distribution of model
explanations, following the predictive, generative, or reinforcement
frameworks, to rapidly derive multiple model explanations. We also analyze the
limitations of an efficient XAI pipeline from the perspectives of the training
phase, the deployment phase, and the use scenarios. Finally, we summarize the
challenges of deploying XAI acceleration methods to real-world scenarios,
overcoming the trade-off between faithfulness and efficiency, and the selection
of different acceleration methods.Comment: 15 pages, 3 figure
Towards Assumption-free Bias Mitigation
Despite the impressive prediction ability, machine learning models show
discrimination towards certain demographics and suffer from unfair prediction
behaviors. To alleviate the discrimination, extensive studies focus on
eliminating the unequal distribution of sensitive attributes via multiple
approaches. However, due to privacy concerns, sensitive attributes are often
either unavailable or missing in real-world scenarios. Therefore, several
existing works alleviate the bias without sensitive attributes. Those studies
face challenges, either in inaccurate predictions of sensitive attributes or
the need to mitigate unequal distribution of manually defined non-sensitive
attributes related to bias. The latter requires strong assumptions about the
correlation between sensitive and non-sensitive attributes. As data
distribution and task goals vary, the strong assumption on non-sensitive
attributes may not be valid and require domain expertise. In this work, we
propose an assumption-free framework to detect the related attributes
automatically by modeling feature interaction for bias mitigation. The proposed
framework aims to mitigate the unfair impact of identified biased feature
interactions. Experimental results on four real-world datasets demonstrate that
our proposed framework can significantly alleviate unfair prediction behaviors
by considering biased feature interactions
DiscoverPath: A Knowledge Refinement and Retrieval System for Interdisciplinarity on Biomedical Research
The exponential growth in scholarly publications necessitates advanced tools
for efficient article retrieval, especially in interdisciplinary fields where
diverse terminologies are used to describe similar research. Traditional
keyword-based search engines often fall short in assisting users who may not be
familiar with specific terminologies. To address this, we present a knowledge
graph-based paper search engine for biomedical research to enhance the user
experience in discovering relevant queries and articles. The system, dubbed
DiscoverPath, employs Named Entity Recognition (NER) and part-of-speech (POS)
tagging to extract terminologies and relationships from article abstracts to
create a KG. To reduce information overload, DiscoverPath presents users with a
focused subgraph containing the queried entity and its neighboring nodes and
incorporates a query recommendation system, enabling users to iteratively
refine their queries. The system is equipped with an accessible Graphical User
Interface that provides an intuitive visualization of the KG, query
recommendations, and detailed article information, enabling efficient article
retrieval, thus fostering interdisciplinary knowledge exploration. DiscoverPath
is open-sourced at https://github.com/ynchuang/DiscoverPath
Recent work on sprite spectrum in Taiwan
campaigns in Taiwan. We first introduce two types of spectroimagers, the slit and slitless types, and discuss their advantages and shortcomings. Next we explore the instrument development and procedures undertaken for this study. In 2006, a slit spectroimager was installed for a sprite campaign and on 15 August of that year, two sprite spectra were recorded using the slit spectroimager along with seven sprites, one halo, one ELVES emission and two jets. By the end of 2015, a slitless spectroimager had been successfully constructed and was ready to conduct additional investigations. On 7 May 2016, a sprite spectrum was recorded using the slitless spectroimager. Following an examination of the calibrations (comprising detection region field of view, wavelength calibration, and response curve), data analysis, and additional calibrations (comprising elevation and azimuthal angles, atmospheric transmittance, and theoretical wavelength calculations) performed in this study, we present the results from our observed sprite spectra using the slit and slitless spectroimagers
An Overview of Regional Experiments on Biomass Burning Aerosols and Related Pollutants in Southeast Asia: From BASE-ASIA and the Dongsha Experiment to 7-SEAS
By modulating the Earth-atmosphere energy, hydrological and biogeochemical cycles, and affecting regional-to-global weather and climate, biomass burning is recognized as one of the major factors affecting the global carbon cycle. However, few comprehensive and wide-ranging experiments have been conducted to characterize biomass-burning pollutants in Southeast Asia (SEA) or assess their regional impact on meteorology, the hydrological cycle, the radiative budget, or climate change. Recently, BASEASIA (Biomass-burning Aerosols in South-East Asia: Smoke Impact Assessment) and the 7-SEAS (7- South-East Asian Studies) Dongsha Experiment were conducted during the spring seasons of 2006 and 2010 in northern SEA, respectively, to characterize the chemical, physical, and radiative properties of biomass-burning emissions near the source regions, and assess their effects. This paper provides an overview of results from these two campaigns and related studies collected in this special issue, entitled Observation, modeling and impact studies of biomass burning and pollution in the SE Asian Environment. This volume includes 28 papers, which provide a synopsis of the experiments, regional weatherclimate, chemical characterization of biomass-burning aerosols and related pollutants in source and sink regions, the spatial distribution of air toxics (atmospheric mercury and dioxins) in source and remote areas, a characterization of aerosol physical, optical, and radiative properties, as well as modeling and impact studies. These studies, taken together, provide the first relatively complete dataset of aerosol chemistry and physical observations conducted in the sourcesink region in the northern SEA, with particular emphasis on the marine boundary layer and lower free troposphere (LFT). The data, analysis and modeling included in these papers advance our present knowledge of source characterization of biomass-burning pollutants near the source regions as well as the physical and chemical processes along transport pathways. In addition, we raise key questions to be addressed by a coming deployment during springtime 2013 in northern SEA, named 7-SEASBASELInE (Biomass-burning Aerosols Stratocumulus Environment: Lifecycles and Interactions Experiment). This campaign will include a synergistic approach for further exploring many key atmospheric processes (e.g., complex aerosol-cloud interactions) and impacts of biomass burning on the surface-atmosphere energy budgets during the lifecycles of biomass burning emissions